query space
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- North America > Canada (0.04)
- (4 more...)
A Discussion of the generative model 1
Thus, we verify that the random effects estimator is equivalent to the generative model (1). Specifically, if u(x) = 1 for all x X, we use ( X, P,ψ) for simplicity. Due to the separability of ψ, we have the following coreset definition. Definitions 2.2 and 2.3, the regression objectives of OLSE and GLSE can be decomposed into Thus, we can apply the above definition to define coresets for OLSE and GLSE. Now we are ready to describe the FL framework in the language of a query space. We first prove Theorem C.1 and propose the corresponding algorithm that constructs an Next, we prove Theorem C.2 and propose the corresponding algorithm that constructs an accurate Caratheodory's Theorem, there must exist at most To accelerate the running time, Jubran et al. [ By the Caratheodory's Theorem, there must exist at most In this section, we complete the proofs for GLSE.
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- (4 more...)
- North America > United States > California (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
A Proof of Lemma 4.4 The first stage of Algorithm 1 outputs an entity level
A.1 The Feldman-Langberg framework We first give the definition of query space and the corresponding coresets. Specifically, if u(x) = 1 for all x X, we use ( X, P,f) for simplicity. Due to the separability of f, we have the following coreset definition. Then by Definition A.2, Lemma 4.4 represents that Now we are ready to give the Feldman-Langberg framework. We also introduce a notion which measures the combinatorial complexity of a query space.
Stronger Coreset Bounds for Kernel Density Estimators via Chaining
Bozzai, Rainie, Rothvoss, Thomas
We apply the discrepancy method and a chaining approach to give improved bounds on the coreset complexity of a wide class of kernel functions. Our results give randomized polynomial time algorithms to produce coresets of size $O\big(\frac{\sqrt{d}}{\varepsilon}\sqrt{\log\log \frac{1}{\varepsilon}}\big)$ for the Gaussian and Laplacian kernels in the case that the data set is uniformly bounded, an improvement that was not possible with previous techniques. We also obtain coresets of size $O\big(\frac{1}{\varepsilon}\sqrt{\log\log \frac{1}{\varepsilon}}\big)$ for the Laplacian kernel for $d$ constant. Finally, we give the best known bounds of $O\big(\frac{\sqrt{d}}{\varepsilon}\sqrt{\log(2\max\{1,\alpha\})}\big)$ on the coreset complexity of the exponential, Hellinger, and JS Kernels, where $1/\alpha$ is the bandwidth parameter of the kernel.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- (3 more...)
Provable Data Subset Selection For Efficient Neural Network Training
Tukan, Murad, Zhou, Samson, Maalouf, Alaa, Rus, Daniela, Braverman, Vladimir, Feldman, Dan
Radial basis function neural networks (\emph{RBFNN}) are {well-known} for their capability to approximate any continuous function on a closed bounded set with arbitrary precision given enough hidden neurons. In this paper, we introduce the first algorithm to construct coresets for \emph{RBFNNs}, i.e., small weighted subsets that approximate the loss of the input data on any radial basis function network and thus approximate any function defined by an \emph{RBFNN} on the larger input data. In particular, we construct coresets for radial basis and Laplacian loss functions. We then use our coresets to obtain a provable data subset selection algorithm for training deep neural networks. Since our coresets approximate every function, they also approximate the gradient of each weight in a neural network, which is a particular function on the input. We then perform empirical evaluations on function approximation and dataset subset selection on popular network architectures and data sets, demonstrating the efficacy and accuracy of our coreset construction.
- North America > United States (0.04)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
Active Learning of Quantum System Hamiltonians yields Query Advantage
Dutt, Arkopal, Pednault, Edwin, Wu, Chai Wah, Sheldon, Sarah, Smolin, John, Bishop, Lev, Chuang, Isaac L.
Hamiltonian learning is an important procedure in quantum system identification, calibration, and successful operation of quantum computers. Through queries to the quantum system, this procedure seeks to obtain the parameters of a given Hamiltonian model and description of noise sources. Standard techniques for Hamiltonian learning require careful design of queries and $O(\epsilon^{-2})$ queries in achieving learning error $\epsilon$ due to the standard quantum limit. With the goal of efficiently and accurately estimating the Hamiltonian parameters within learning error $\epsilon$ through minimal queries, we introduce an active learner that is given an initial set of training examples and the ability to interactively query the quantum system to generate new training data. We formally specify and experimentally assess the performance of this Hamiltonian active learning (HAL) algorithm for learning the six parameters of a two-qubit cross-resonance Hamiltonian on four different superconducting IBM Quantum devices. Compared with standard techniques for the same problem and a specified learning error, HAL achieves up to a $99.8\%$ reduction in queries required, and a $99.1\%$ reduction over the comparable non-adaptive learning algorithm. Moreover, with access to prior information on a subset of Hamiltonian parameters and given the ability to select queries with linearly (or exponentially) longer system interaction times during learning, HAL can exceed the standard quantum limit and achieve Heisenberg (or super-Heisenberg) limited convergence rates during learning.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (4 more...)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Uncertainty > Bayesian Inference (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Inductive Learning (0.86)